national laboratory
Leveraging AI for Productive and Trustworthy HPC Software: Challenges and Research Directions
Teranishi, Keita, Menon, Harshitha, Godoy, William F., Balaprakash, Prasanna, Bau, David, Ben-Nun, Tal, Bhatele, Abhinav, Franchetti, Franz, Franusich, Michael, Gamblin, Todd, Georgakoudis, Giorgis, Goldstein, Tom, Guha, Arjun, Hahn, Steven, Iancu, Costin, Jin, Zheming, Jones, Terry, Low, Tze Meng, Mankad, Het, Miniskar, Narasinga Rao, Monil, Mohammad Alaul Haque, Nichols, Daniel, Parasyris, Konstantinos, Pophale, Swaroop, Valero-Lara, Pedro, Vetter, Jeffrey S., Williams, Samuel, Young, Aaron
We discuss the challenges and propose research directions for using AI to revolutionize the development of high-performance computing (HPC) software. AI technologies, in particular large language models, have transformed every aspect of software development. For its part, HPC software is recognized as a highly specialized scientific field of its own. We discuss the challenges associated with leveraging state-of-the-art AI technologies to develop such a unique and niche class of software and outline our research directions in the two US Department of Energy--funded projects for advancing HPC Software via AI: Ellora and Durban.
- North America > United States > Maryland > Prince George's County > College Park (0.14)
- North America > United States > New York > New York County > New York City (0.05)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- (6 more...)
- Government > Regional Government > North America Government > United States Government (1.00)
- Energy (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Why the AI Industry Is Betting on a Fusion Energy Breakthrough
Booth is a reporter at TIME. Booth is a reporter at TIME. When Sam Altman arrived at Helion Energy's small Redmond, Wash., office in early 2014, nuclear-fusion textbooks tucked under his arm, the company was focusing its efforts on research and development. By the time he left, several days later, he had persuaded the fusion-energy startup to chart a more aggressive path toward deployment, CEO David Kirtley recalls. A year later, Altman, who was co-founding OpenAI around the same time, invested $9.5 million in Helion, taking the role of chairman.
- North America > United States > Washington > King County > Redmond (0.25)
- Oceania > New Zealand (0.05)
- North America > United States > California (0.05)
- Asia > China (0.05)
- Information Technology (1.00)
- Energy > Power Industry > Utilities > Nuclear (1.00)
Extrapolating Phase-Field Simulations in Space and Time with Purely Convolutional Architectures
Bonneville, Christophe, Bieberdorf, Nathan, Robbe, Pieterjan, Asta, Mark, Najm, Habib N., Capolungo, Laurent, Safta, Cosmin
Phase-field models of liquid metal dealloying (LMD) can resolve rich microstructural dynamics but become intractable for large domains or long time horizons. We present a conditionally parameterized, fully convolutional U-Net surrogate that generalizes far beyond its training window in both space and time. The design integrates convolutional self-attention and physics-aware padding, while parameter conditioning enables variable time-step skipping and adaptation to diverse alloy systems. Although trained only on short, small-scale simulations, the surrogate exploits the translational invariance of convolutions to extend predictions to much longer horizons than traditional solvers. It accurately reproduces key LMD physics, with relative errors typically under 5% within the training regime and below 10% when extrapolating to larger domains and later times. The method accelerates computations by up to 16,000 times, cutting weeks of simulation down to seconds, and marks an early step toward scalable, high-fidelity extrapolation of LMD phase-field models.
- North America > United States > California > Alameda County > Berkeley (0.14)
- Europe > Switzerland > Zürich > Zürich (0.14)
- North America > United States > New Mexico > Los Alamos County > Los Alamos (0.05)
- North America > United States > California > Alameda County > Livermore (0.04)
- Government > Regional Government > North America Government > United States Government (1.00)
- Energy (1.00)
CelloAI: Leveraging Large Language Models for HPC Software Development in High Energy Physics
Atif, Mohammad, Chopra, Kriti, Kilic, Ozgur, Wang, Tianle, Dong, Zhihua, Leggett, Charles, Lin, Meifeng, Calafiura, Paolo, Habib, Salman
Next-generation High Energy Physics (HEP) experiments will generate unprecedented data volumes, necessitating High Performance Computing (HPC) integration alongside traditional high-throughput computing. However, HPC adoption in HEP is hindered by the challenge of porting legacy software to heterogeneous architectures and the sparse documentation of these complex scientific codebases. We present CelloAI, a locally hosted coding assistant that leverages Large Language Models (LLMs) with retrieval-augmented generation (RAG) to support HEP code documentation and generation. This local deployment ensures data privacy, eliminates recurring costs and provides access to large context windows without external dependencies. CelloAI addresses two primary use cases, code documentation and code generation, through specialized components. For code documentation, the assistant provides: (a) Doxygen style comment generation for all functions and classes by retrieving relevant information from RAG sources (papers, posters, presentations), (b) file-level summary generation, and (c) an interactive chatbot for code comprehension queries. For code generation, CelloAI employs syntax-aware chunking strategies that preserve syntactic boundaries during embedding, improving retrieval accuracy in large codebases. The system integrates callgraph knowledge to maintain dependency awareness during code modifications and provides AI-generated suggestions for performance optimization and accurate refactoring. We evaluate CelloAI using real-world HEP applications from ATLAS, CMS, and DUNE experiments, comparing different embedding models for code retrieval effectiveness. Our results demonstrate the AI assistant's capability to enhance code understanding and support reliable code generation while maintaining the transparency and safety requirements essential for scientific computing environments.
- Information Technology > Security & Privacy (1.00)
- Energy (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Machine Learning-driven Multiscale MD Workflows: The Mini-MuMMI Experience
Pottier, Loïc, Georgouli, Konstantia, Carpenter, Timothy S., Aydin, Fikret, Tempkin, Jeremy O. B., Nissley, Dwight V., Streitz, Frederick H., Scogland, Thomas R. W., Bremer, Peer-Timo, Lightstone, Felice C., Ingólfsson, Helgi I.
Computational models have become one of the prevalent methods to model complex phenomena. To accurately model complex interactions, such as detailed biomolecular interactions, scientists often rely on multiscale models comprised of several internal models operating at difference scales, ranging from microscopic to macroscopic length and time scales. Bridging the gap between different time and length scales has historically been challenging but the advent of newer machine learning (ML) approaches has shown promise for tackling that task. Multiscale models require massive amounts of computational power and a powerful workflow management system. Orchestrating ML-driven multiscale studies on parallel systems with thousands of nodes is challenging, the workflow must schedule, allocate and control thousands of simulations operating at different scales. Here, we discuss the massively parallel Multiscale Machine-Learned Modeling Infrastructure (MuMMI), a multiscale workflow management infrastructure, that can orchestrate thousands of molecular dynamics (MD) simulations operating at different timescales, spanning from millisecond to nanosecond. More specifically, we introduce a novel version of MuMMI called "mini-MuMMI". Mini-MuMMI is a curated version of MuMMI designed to run on modest HPC systems or even laptops whereas MuMMI requires larger HPC systems. We demonstrate mini-MuMMI utility by exploring RAS-RAF membrane interactions and discuss the different challenges behind the generalization of multiscale workflows and how mini-MuMMI can be leveraged to target a broader range of applications outside of MD and RAS-RAF interactions.
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- North America > United States > Tennessee > Anderson County > Oak Ridge (0.04)
- North America > United States > New Mexico > Bernalillo County > Albuquerque (0.04)
- (2 more...)
- Workflow (1.00)
- Research Report (0.82)
An Active Learning-Based Streaming Pipeline for Reduced Data Training of Structure Finding Models in Neutron Diffractometry
Wang, Tianle, Ramirez, Jorge, Garcia-Cardona, Cristina, Proffen, Thomas, Jha, Shantenu, Seal, Sudip K.
Structure determination workloads in neutron diffractometry are computationally expensive and routinely require several hours to many days to determine the structure of a material from its neutron diffraction patterns. The potential for machine learning models trained on simulated neutron scattering patterns to significantly speed up these tasks have been reported recently. However, the amount of simulated data needed to train these models grows exponentially with the number of structural parameters to be predicted and poses a significant computational challenge. To overcome this challenge, we introduce a novel batch-mode active learning (AL) policy that uses uncertainty sampling to simulate training data drawn from a probability distribution that prefers labelled examples about which the model is least certain. We confirm its efficacy in training the same models with about 75% less training data while improving the accuracy. We then discuss the design of an efficient stream-based training workflow that uses this AL policy and present a performance study on two heterogeneous platforms to demonstrate that, compared with a conventional training workflow, the streaming workflow delivers about 20% shorter training time without any loss of accuracy.
- North America > United States > New Mexico > Los Alamos County > Los Alamos (0.04)
- North America > United States > New Jersey (0.04)
- Research Report (1.00)
- Workflow (0.89)
- Energy (1.00)
- Government > Regional Government > North America Government > United States Government (0.94)
AstroMLab 4: Benchmark-Topping Performance in Astronomy Q&A with a 70B-Parameter Domain-Specialized Reasoning Model
de Haan, Tijmen, Ting, Yuan-Sen, Ghosal, Tirthankar, Nguyen, Tuan Dung, Accomazzi, Alberto, Herron, Emily, Lama, Vanessa, Pan, Rui, Wells, Azton, Ramachandra, Nesar
General-purpose large language models, despite their broad capabilities, often struggle with specialized domain knowledge, a limitation particularly pronounced in more accessible, lower-parameter versions. This gap hinders their deployment as effective agents in demanding fields such as astronomy. Building on our prior work with AstroSage-8B, this study introduces AstroSage-70B, a significantly larger and more advanced domain-specialized natural-language AI assistant. It is designed for research and education across astronomy, astrophysics, space science, astroparticle physics, cosmology, and astronomical instrumentation. Developed from the Llama-3.1-70B foundation, AstroSage-70B underwent extensive continued pre-training on a vast corpus of astronomical literature, followed by supervised fine-tuning and model merging. Beyond its 70-billion parameter scale, this model incorporates refined datasets, judiciously chosen learning hyperparameters, and improved training procedures, achieving state-of-the-art performance on complex astronomical tasks. Notably, we integrated reasoning chains into the SFT dataset, enabling AstroSage-70B to either answer the user query immediately, or first emit a human-readable thought process. Evaluated on the AstroMLab-1 benchmark -- comprising 4,425 questions from literature withheld during training -- AstroSage-70B achieves state-of-the-art performance. It surpasses all other tested open-weight and proprietary models, including leading systems like o3, Gemini-2.5-Pro, Claude-3.7-Sonnet, Deepseek-R1, and Qwen-3-235B, even those with API costs two orders of magnitude higher. This work demonstrates that domain specialization, when applied to large-scale models, can enable them to outperform generalist counterparts in specialized knowledge areas like astronomy, thereby advancing the frontier of AI capabilities in the field.
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.14)
- North America > United States > Ohio > Franklin County > Columbus (0.04)
- Asia > Japan > Honshū > Kantō > Ibaraki Prefecture > Tsukuba (0.04)
- (6 more...)
- Research Report (1.00)
- Personal > Interview (0.40)
- Energy (0.69)
- Government > Regional Government > North America Government > United States Government (0.69)
- Education (0.68)
Generalizable Implicit Neural Representations via Parameterized Latent Dynamics for Baroclinic Ocean Forecasting
Zhao, Guang, Luo, Xihaier, Lee, Seungjun, Ren, Yihui, Yoo, Shinjae, Van Roekel, Luke, Nadiga, Balu, Narayanan, Sri Hari Krishna, Sun, Yixuan, Xu, Wei
Published as a workshop paper at "Tackling Climate Change with Machine Learning", ICLR 2025 Mesoscale ocean dynamics play a critical role in climate systems, governing heat transport, hurricane genesis, and drought patterns. However, simulating these processes at high resolution remains computationally prohibitive due to their nonlinear, multiscale nature and vast spatiotemporal domains. Implicit neural representations (INRs) reduce the computational costs as resolution-independent surrogates but fail in many-query scenarios (inverse modeling) requiring rapid evaluations across diverse parameters. We present PINROD, a novel framework combining dynamics-aware implicit neural representations with parametrized neural ordinary differential equations to address these limitations. Experiments on ocean mesoscale activity data show superior accuracy over existing baselines and improved computational efficiency compared to standard numerical simulations.
- North America > United States > New Mexico > Los Alamos County > Los Alamos (0.05)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Energy (1.00)
- Government > Regional Government (0.47)
EAIRA: Establishing a Methodology for Evaluating AI Models as Scientific Research Assistants
Cappello, Franck, Madireddy, Sandeep, Underwood, Robert, Getty, Neil, Chia, Nicholas Lee-Ping, Ramachandra, Nesar, Nguyen, Josh, Keceli, Murat, Mallick, Tanwi, Li, Zilinghan, Ngom, Marieme, Zhang, Chenhui, Yanguas-Gil, Angel, Antoniuk, Evan, Kailkhura, Bhavya, Tian, Minyang, Du, Yufeng, Ting, Yuan-Sen, Wells, Azton, Nicolae, Bogdan, Maurya, Avinash, Rafique, M. Mustafa, Huerta, Eliu, Li, Bo, Foster, Ian, Stevens, Rick
Recent advancements have positioned AI, and particularly Large Language Models (LLMs), as transformative tools for scientific research, capable of addressing complex tasks that require reasoning, problem-solving, and decision-making. Their exceptional capabilities suggest their potential as scientific research assistants but also highlight the need for holistic, rigorous, and domain-specific evaluation to assess effectiveness in real-world scientific applications. This paper describes a multifaceted methodology for Evaluating AI models as scientific Research Assistants (EAIRA) developed at Argonne National Laboratory. This methodology incorporates four primary classes of evaluations. 1) Multiple Choice Questions to assess factual recall; 2) Open Response to evaluate advanced reasoning and problem-solving skills; 3) Lab-Style Experiments involving detailed analysis of capabilities as research assistants in controlled environments; and 4) Field-Style Experiments to capture researcher-LLM interactions at scale in a wide range of scientific domains and applications. These complementary methods enable a comprehensive analysis of LLM strengths and weaknesses with respect to their scientific knowledge, reasoning abilities, and adaptability. Recognizing the rapid pace of LLM advancements, we designed the methodology to evolve and adapt so as to ensure its continued relevance and applicability. This paper describes the methodology state at the end of February 2025. Although developed within a subset of scientific domains, the methodology is designed to be generalizable to a wide range of scientific domains.
- North America > United States > Pennsylvania (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > New Mexico > Los Alamos County > Los Alamos (0.04)
- (7 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.67)
- Health & Medicine (1.00)
- Energy (1.00)
- Education (1.00)
- Government > Regional Government > North America Government > United States Government (0.93)
AI-powered supercomputer to start testing America's nukes
Scientists have unveiled the world's fastest computer in California that will be used to secure America's nuclear weapons stockpile. The 600 million exascale supercomputer, called'El Capitan,' is only the third of its kind in the world. That's equivalent to the processing power of about one million high-end smartphones working simultaneously, researchers said. El Capitan launched at the Livermore National Laboratory (LNNL) in November 2024 and was officially announced to the public on January 9. It will primarily focus on national security, including nuclear data and weapon testing, high-energy-density physics, materials discovery and other sensitive or classified tasks.
- North America > United States > California (0.27)
- North America > United States > New Mexico > Los Alamos County > Los Alamos (0.05)